[ 
https://issues.apache.org/jira/browse/PIG-4937?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15351312#comment-15351312
 ] 

liyunzhang_intel commented on PIG-4937:
---------------------------------------

[~rohini]:  As there are 3 nodes( each has 56g) and totally 168(56*3) g.  As i 
assign mapreduce.reduce.memory.mb as 2g and set mappers(configure the number of 
mappers for data generator) from 90 to 40 because 90*2=180g is over 168g then 
there will be not enough memory for resourceManager.

mapred-site.xml
{code}
<configuration>

  <property>
    <name>mapreduce.framework.name</name>
    <value>yarn</value>
  </property>
 <property>
    <name>mapred.child.java.opts</name>
    <value>-Xmx1024m</value>
  </property>
  <property>
    <name>mapreduce.map.memory.mb</name>
    <value>2048</value>
  </property>
  <property>
    <name>mapreduce.reduce.memory.mb</name>
    <value>2048</value>
  </property>

  <property>
    <name>mapreduce.task.io.sort.mb</name>
    <value>1024</value>
  </property>
</configuration>

{code}


Now it shows that 40 mappers are still running(see attached pigmix1.png) and 
the status of a single map task is 4800000(see attached pigmix2.png).   It is 
supposed that every map task will generate 625000000/40=15625000 (detailed see 
[DataGenerator|https://github.com/apache/pig/blob/trunk/test/perf/pigmix/src/java/org/apache/pig/test/pigmix/datagen/DataGenerator.java#L816]).
 Is my understanding right?

> Pigmix hangs when generating data after rows  is set as 625000000 in  
> test/perf/pigmix/conf/config.sh
> -----------------------------------------------------------------------------------------------------
>
>                 Key: PIG-4937
>                 URL: https://issues.apache.org/jira/browse/PIG-4937
>             Project: Pig
>          Issue Type: Bug
>            Reporter: liyunzhang_intel
>         Attachments: pigmix1.PNG
>
>
> use the default setting in test/perf/pigmix/conf/config.sh, generate data by
> "ant -v -Dharness.hadoop.home=$HADOOP_HOME -Dhadoopversion=23  pigmix-deploy 
> >ant.pigmix.deploy"
> it hangs in the log:
> {code}
>  [exec] Generating mapping file for column d:1:100000:z:5 into 
> hdfs://bdpe41:8020/user/root/tmp/tmp-1056793210/tmp-786100428
>      [exec] processed 99%.
>      [exec] Generating input files into 
> hdfs://bdpe41:8020/user/root/tmp/tmp-1056793210/tmp595036324
>      [exec] Submit hadoop job...
>      [exec] 16/06/25 23:06:32 INFO client.RMProxy: Connecting to 
> ResourceManager at bdpe41/10.239.47.137:8032
>      [exec] 16/06/25 23:06:32 INFO client.RMProxy: Connecting to 
> ResourceManager at bdpe41/10.239.47.137:8032
>      [exec] 16/06/25 23:06:32 INFO mapred.FileInputFormat: Total input paths 
> to process : 90
>      [exec] 16/06/25 23:06:32 INFO mapreduce.JobSubmitter: number of splits:90
>      [exec] 16/06/25 23:06:32 INFO mapreduce.JobSubmitter: Submitting tokens 
> for job: job_1466776148247_0034
>      [exec] 16/06/25 23:06:33 INFO impl.YarnClientImpl: Submitted application 
> application_1466776148247_0034
>      [exec] 16/06/25 23:06:33 INFO mapreduce.Job: The url to track the job: 
> http://bdpe41:8088/proxy/application_1466776148247_0034/     [exec] 16/06/25 
> 23:06:33 INFO mapreduce.Job: Running job: job_1466776148247_0034
>      [exec] 16/06/25 23:06:38 INFO mapreduce.Job: Job job_1466776148247_0034 
> running in uber mode : false
>      [exec] 16/06/25 23:06:38 INFO mapreduce.Job:  map 0% reduce 0%
>      [exec] 16/06/25 23:06:53 INFO mapreduce.Job:  map 2% reduce 0%
>      [exec] 16/06/25 23:06:59 INFO mapreduce.Job:  map 26% reduce 0%
>      [exec] 16/06/25 23:07:00 INFO mapreduce.Job:  map 61% reduce 0%
>      [exec] 16/06/25 23:07:02 INFO mapreduce.Job:  map 62% reduce 0%
>      [exec] 16/06/25 23:07:03 INFO mapreduce.Job:  map 64% reduce 0%
>      [exec] 16/06/25 23:07:04 INFO mapreduce.Job:  map 79% reduce 0%
>      [exec] 16/06/25 23:07:05 INFO mapreduce.Job:  map 86% reduce 0%
>      [exec] 16/06/25 23:07:06 INFO mapreduce.Job:  map 92% reduce 0%
> {code}
> When i use 625000 as the rows in test/perf/pigmix/conf/config.sh, it is 
> successful to generate test data. So is the problem on the limit 
> resources(disk size or others)?  My env is 3 nodes cluster(each node has 
> about a disk(830G)) and i assign memory and cpu in the yarn-site.xml like 
> following:
> {code}
>  yarn.nodemanager.resource.memory-mb=56G
>  yarn.nodemanger.resource.cpu-vcores=28
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to